Multimodal End-to-End Autonomous Driving

نویسندگان

چکیده

A crucial component of an autonomous vehicle (AV) is the artificial intelligence (AI) able to drive towards a desired destination. Today, there are different paradigms addressing development AI drivers. On one hand, we find modular pipelines, which divide driving task into sub-tasks such as perception and maneuver planning control. other end-to-end approaches that try learn direct mapping from input raw sensor data control signals. The later relatively less studied, but gaining popularity since they demanding in terms annotation. This paper focuses on driving. So far, most proposals relying this paradigm assume RGB images data. However, AVs will not be equipped only with cameras, also active sensors providing accurate depth information ( e.g. , LiDARs). Accordingly, analyses whether combining modalities, xmlns:xlink="http://www.w3.org/1999/xlink">i.e. using RGBD data, produces better drivers than single modality. We consider multimodality based early, mid late fusion schemes, both multisensory single-sensor (monocular estimation) settings. Using CARLA simulator conditional imitation learning (CIL), show how, indeed, early outperforms single-modality.

برای دانلود رایگان متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Query-Efficient Imitation Learning for End-to-End Autonomous Driving

One way to approach end-to-end autonomous driving is to learn a policy function that maps from a sensory input, such as an image frame from a front-facing camera, to a driving action, by imitating an expert driver, or a reference policy. This can be done by supervised learning, where a policy function is tuned to minimize the difference between the predicted and ground-truth actions. A policy f...

متن کامل

Learning End-to-end Multimodal Sensor Policies for Autonomous Navigation

Sensor fusion is indispensable to improve accuracy and robustness in an autonomous navigation setting. However, in the space of end-to-end sensorimotor control, this multimodal outlook has received limited attention. In this work, we propose a novel stochastic regularization technique, called Sensor Dropout, to robustify multimodal sensor policy learning outcomes. We also introduce an auxiliary...

متن کامل

DDD17: End-To-End DAVIS Driving Dataset

Event cameras, such as dynamic vision sensors (DVS), and dynamic and activepixel vision sensors (DAVIS) can supplement other autonomous driving sensors by providing a concurrent stream of standard active pixel sensor (APS) images and DVS temporal contrast events. The APS stream is a sequence of standard grayscale global-shutter image sensor frames. The DVS events represent brightness changes oc...

متن کامل

Agile Off-Road Autonomous Driving Using End-to-End Deep Imitation Learning

We present an end-to-end imitation learning system for agile, off-road autonomous driving using only low-cost on-board sensors. By imitating a model predictive controller equipped with advanced sensors, we train a deep neural network control policy to map raw, high-dimensional observations to continuous steering and throttle commands. Compared with recent approaches to similar tasks, our method...

متن کامل

End-to-end esophagojejunostomy versus standard end-to-side esophagojejunostomy: which one is preferable?

 Abstract Background: End-to-side esophagojejunostomy has almost always been associated with some degree of dysphagia. To overcome this complication we decided to perform an end-to-end anastomosis and compare it with end-to-side Roux-en-Y esophagojejunostomy. Methods: In this prospective study, between 1998 and 2005, 71 patients with a diagnosis of gastric adenocarcinoma underwent total gastrec...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: IEEE Transactions on Intelligent Transportation Systems

سال: 2022

ISSN: ['1558-0016', '1524-9050']

DOI: https://doi.org/10.1109/tits.2020.3013234